On a Frank-Wolfe type theorem in cubic optimization
نویسندگان
چکیده
منابع مشابه
On the Extensions of Frank - Wolfe Theorem
In this paper we consider optimization problems de ned by a quadratic objective function and a nite number of quadratic inequality constraints. Given that the objective function is bounded over the feasible set, we present a comprehensive study of the conditions under which the optimal solution set is nonempty, thus extending the so-called Frank-Wolfe theorem. In particular, we rst prove a gene...
متن کاملA Frank-Wolfe Type Theorem for Convex Polynomial Programs
In 1956, Frank and Wolfe extended the fundamental existence theorem of linear programming by proving that an arbitrary quadratic function f attains its minimum over a nonempty convex polyhedral set X provided f is bounded from below over X . We show that a similar statement holds if f is a convex polynomial and X is the solution set of a system of convex polynomial inequalities. In fact, this r...
متن کاملA Conic Duality Frank-Wolfe-Type Theorem via Exact Penalization in Quadratic Optimization
The famous Frank–Wolfe theorem ensures attainability of the optimal value for quadratic objective functions over a (possibly unbounded) polyhedron if the feasible values are bounded. This theorem does not hold in general for conic programs where linear constraints are replaced by more general convex constraints like positive-semidefiniteness or copositivity conditions, despite the fact that the...
متن کاملBandit Optimization with Upper-Confidence Frank-Wolfe
We consider the problem of bandit optimization, inspired by stochastic optimization and online learning problems with bandit feedback. In this problem, the objective is to minimize a global loss function of all the actions, not necessarily a cumulative loss. This framework allows us to study a very general class of problems, with applications in statistics, machine learning, and other fields. T...
متن کاملOn the Global Linear Convergence of Frank-Wolfe Optimization Variants
The Frank-Wolfe (FW) optimization algorithm has lately re-gained popularity thanks in particular to its ability to nicely handle the structured constraints appearing in machine learning applications. However, its convergence rate is known to be slow (sublinear) when the solution lies at the boundary. A simple lessknown fix is to add the possibility to take ‘away steps’ during optimization, an o...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Optimization
سال: 2019
ISSN: 0233-1934,1029-4945
DOI: 10.1080/02331934.2019.1566327